6 research outputs found

    Status and prediction of ozone as an air pollutant in Ahvaz City, Iran

    Get PDF
    In the present study, air quality analyses for ozone (O3) were conducted in Ahvaz, a city in the south of Iran. The measurements were taken from 2009 through 2010 in two different locations to prepare average data in the city. Relations between the air pollutant and some meteorological parameters were calculated statistically using the daily average data. The wind data (velocity, direction), relative humidity, temperature, sunshine hours, evaporation and rainfall were considered as independent variables. The relationships between concentration of pollutant and meteorological parameters were expressed by multiple linear and nonlinear regression equations for both annual and seasonal conditions using SPSS software. RMSE test showed that among different prediction model, stepwise model is the best option. The average concentrations were calculated for every 24 hours, each month and each season. Results showed that the highest concentration of ozone occurs generally in the afternoon, while the least concentration is found at the beginning of the morning. Monthly concentrations of ozone showed the highest value in August, while the least value was found in October. The seasonal concentrations showed the highest amounts in summer

    Owl and Lizard: Patterns of Head Pose and Eye Pose in Driver Gaze Classification

    Full text link
    Accurate, robust, inexpensive gaze tracking in the car can help keep a driver safe by facilitating the more effective study of how to improve (1) vehicle interfaces and (2) the design of future Advanced Driver Assistance Systems. In this paper, we estimate head pose and eye pose from monocular video using methods developed extensively in prior work and ask two new interesting questions. First, how much better can we classify driver gaze using head and eye pose versus just using head pose? Second, are there individual-specific gaze strategies that strongly correlate with how much gaze classification improves with the addition of eye pose information? We answer these questions by evaluating data drawn from an on-road study of 40 drivers. The main insight of the paper is conveyed through the analogy of an "owl" and "lizard" which describes the degree to which the eyes and the head move when shifting gaze. When the head moves a lot ("owl"), not much classification improvement is attained by estimating eye pose on top of head pose. On the other hand, when the head stays still and only the eyes move ("lizard"), classification accuracy increases significantly from adding in eye pose. We characterize how that accuracy varies between people, gaze strategies, and gaze regions.Comment: Accepted for Publication in IET Computer Vision. arXiv admin note: text overlap with arXiv:1507.0476
    corecore